34 research outputs found

    A New Forensic Video Database for Source Smartphone Identification: Description and Analysis

    Get PDF
    In recent years, the field of digital imaging has made significant progress, so that today every smartphone has a built-in video camera that allows you to record high-quality video for free and without restrictions. On the other hand, rapidly growing internet technology has contributed significantly to the widespread use of digital video via web-based multimedia systems and mobile smartphone applications such as YouTube, Facebook, Twitter, WhatsApp, etc. However, as the recording and distribution of digital videos have become affordable nowadays, security issues have become threatening and spread worldwide. One of the security issues is identifying source cameras on videos. There are some new challenges that should be addressed in this area. One of the new challenges is individual source camera identification (ISCI), which focuses on identifying each device regardless of its model. The first step towards solving the problems is a popular video database recorded by modern smartphone devices, which can also be used for deep learning methods that are growing rapidly in the field of source camera identification. In this paper, a smartphone video database named Qatar University Forensic Video Database (QUFVD) is introduced. The QUFVD includes 6000 videos from 20 modern smartphone representing five brands, each brand has two models, and each model has two identical smartphone devices. This database is suitable for evaluating different techniques such as deep learning methods for video source smartphone identification and verification. To evaluate the QUFVD, a series of experiments to identify source cameras using a deep learning technique are conducted. The results show that improvements are essential for the ISCI scenario on video

    Video surveillance using deep transfer learning and deep domain adaptation: Towards better generalization

    Get PDF
    Recently, developing automated video surveillance systems (VSSs) has become crucial to ensure the security and safety of the population, especially during events involving large crowds, such as sporting events. While artificial intelligence (AI) smooths the path of computers to think like humans, machine learning (ML) and deep learning (DL) pave the way more, even by adding training and learning components. DL algorithms require data labeling and high-performance computers to effectively analyze and understand surveillance data recorded from fixed or mobile cameras installed in indoor or outdoor environments. However, they might not perform as expected, take much time in training, or not have enough input data to generalize well. To that end, deep transfer learning (DTL) and deep domain adaptation (DDA) have recently been proposed as promising solutions to alleviate these issues. Typically, they can (i) ease the training process, (ii) improve the generalizability of ML and DL models, and (iii) overcome data scarcity problems by transferring knowledge from one domain to another or from one task to another. Although the increasing number of articles proposed to develop DTL- and DDA-based VSSs, a thorough review that summarizes and criticizes the state-of-the-art is still missing. To that end, this paper introduces, to the best of the authors' knowledge, the first overview of existing DTL- and DDA-based video surveillance to (i) shed light on their benefits, (ii) discuss their challenges, and (iii) highlight their future perspectives.This research work was made possible by research grant support (QUEX-CENG-SCDL-19/20-1) from Supreme Committee for Delivery and Legacy (SC) in Qatar. The statements made herein are solely the responsibility of the authors. Open Access funding provided by the Qatar National Library.Scopu

    Gait recognition for person re-identification

    Get PDF
    Person re-identification across multiple cameras is an essential task in computer vision applications, particularly tracking the same person in different scenes. Gait recognition, which is the recognition based on the walking style, is mostly used for this purpose due to that human gait has unique characteristics that allow recognizing a person from a distance. However, human recognition via gait technique could be limited with the position of captured images or videos. Hence, this paper proposes a gait recognition approach for person re-identification. The proposed approach starts with estimating the angle of the gait first, and this is then followed with the recognition process, which is performed using convolutional neural networks. Herein, multitask convolutional neural network models and extracted gait energy images (GEIs) are used to estimate the angle and recognize the gait. GEIs are extracted by first detecting the moving objects, using background subtraction techniques. Training and testing phases are applied to the following three recognized datasets: CASIA-(B), OU-ISIR, and OU-MVLP. The proposed method is evaluated for background modeling using the Scene Background Modeling and Initialization (SBI) dataset. The proposed gait recognition method showed an accuracy of more than 98% for almost all datasets. Results of the proposed approach showed higher accuracy compared to obtained results of other methods result for CASIA-(B) and OU-MVLP and form the best results for the OU-ISIR dataset

    3D objects and scenes classification, recognition, segmentation, and reconstruction using 3D point cloud data: A review

    Full text link
    Three-dimensional (3D) point cloud analysis has become one of the attractive subjects in realistic imaging and machine visions due to its simplicity, flexibility and powerful capacity of visualization. Actually, the representation of scenes and buildings using 3D shapes and formats leveraged many applications among which automatic driving, scenes and objects reconstruction, etc. Nevertheless, working with this emerging type of data has been a challenging task for objects representation, scenes recognition, segmentation, and reconstruction. In this regard, a significant effort has recently been devoted to developing novel strategies, using different techniques such as deep learning models. To that end, we present in this paper a comprehensive review of existing tasks on 3D point cloud: a well-defined taxonomy of existing techniques is performed based on the nature of the adopted algorithms, application scenarios, and main objectives. Various tasks performed on 3D point could data are investigated, including objects and scenes detection, recognition, segmentation and reconstruction. In addition, we introduce a list of used datasets, we discuss respective evaluation metrics and we compare the performance of existing solutions to better inform the state-of-the-art and identify their limitations and strengths. Lastly, we elaborate on current challenges facing the subject of technology and future trends attracting considerable interest, which could be a starting point for upcoming research studie

    PRNU-Net: a Deep Learning Approach for Source Camera Model Identification based on Videos Taken with Smartphone

    Get PDF
    Recent advances in digital imaging have meant that every smartphone has a video camera that can record highquality video for free and without restrictions. In addition, rapidly developing Internet technology has contributed significantly to the widespread distribution of digital video via web-based multimedia systems and mobile applications such as YouTube, Facebook, Twitter, WhatsApp, etc. However, as the recording and distribution of digital video has become affordable nowadays, security issues have become threatening and have spread worldwide. One of the security issues is the identification of source cameras on videos. Generally, two common categories of methods are used in this area, namely Photo Response Non-Uniformity (PRNU) and Machine Learning approaches. To exploit the power of both approaches, this work adds a new PRNU-based layer to a convolutional neural network (CNN) called PRNU-Net. To explore the new layer, the main structure of the CNN is based on the MISLnet, which has been used in several studies to identify the source camera. The experimental results show that the PRNU-Net is more successful than the MISLnet and that the PRNU extracted by the layer from low features, namely edges or textures, is more useful than high and mid-level features, namely parts and objects, in classifying source camera models. On average, the network improves theresults in a new database by about 4

    A combined multiple action recognition and summarization for surveillance video sequences

    Get PDF
    Human action recognition and video summarization represent challenging tasks for several computer vision applications including video surveillance, criminal investigations, and sports applications. For long videos, it is difficult to search within a video for a specific action and/or person. Usually, human action recognition approaches presented in the literature deal with videos that contain only a single person, and they are able to recognize his action. This paper proposes an effective approach to multiple human action detection, recognition, and summarization. The multiple action detection extracts human bodies’ silhouette, then generates a specific sequence for each one of them using motion detection and tracking method. Each of the extracted sequences is then divided into shots that represent homogeneous actions in the sequence using the similarity between each pair frames. Using the histogram of the oriented gradient (HOG) of the Temporal Difference Map (TDMap) of the frames of each shot, we recognize the action by performing a comparison between the generated HOG and the existed HOGs in the training phase which represents all the HOGs of many actions using a set of videos for training. Also, using the TDMap images we recognize the action using a proposed CNN model. Action summarization is performed for each detected person. The efficiency of the proposed approach is shown through the obtained results for mainly multi-action detection and recognition

    Self-sanitizing reusable glove via 3D-printing and common mold making method

    Get PDF
    In health care and public health practice, it is critical to settings control practices that are critical to reducing the transmission of infections through cross-contamination. To provide protection from cross-contamination, use and throw gloves are routinely used. However, single-time use and inconsistent sanitization of used gloves remain a large problem and elevate the risk of catching viruses, germs, pathogens, and contaminants. The study reports reusable self-sanitizing gloves via 3D-printing and common hand molding methods. The major contribution is frequent self-sanitization of gloves without any manual intervention. The elastomeric material is used for fabricating gloves and continuous channels are embedded within the elastomeric material that runs through the entire glove surface, covering the front, back, and fingers. Elastomeric material allows the engagement of fingers for gripping objects. While the embedded channel is provided with uniformly spaced openings to eject the sanitizing solution. The glove surface is textured with a porous morphology that acts as mini and micro reservoirs for sterilizing solution ejected through embedded channel opening. The embedded channel is connected to a sanitizing solution storage tank. The incorporation of sanitizing solution storage tank enables its usage over a longer period. This uniquely constructed design of the gloves even assists in the effective sterilization of infected surface that comes in contact with the gloves. The gloves can be customized to improve comfortability by fabricating them from the 3D-printed mound developed based on the palm size of the user. The developed technology can be used by individuals working in hospitals, the transport sector, delivery units, schools, offices, industries, etc. We strongly believe that this technology will be highly useful in minimizing the risk of getting infected through cross-contamination and will help in maintaining hygienic as well as safe surroundings.This work was supported by the RRC-2-063-133 grant from the Qatar National Research Fund (a member of Qatar Foundation). Open Access funding was provided by the Qatar National Library

    Video summarization based on motion detection for surveillance systems

    No full text
    In this paper a video summarization method based on motion detection has been proposed. Sensor noise (noise of acquisition and digitization) and the illumination changes in the scene are the most limitations of the background subtraction approaches. In order to handle these problems, this paper present an approach based on the combining of the background subtraction and the Structure-Texture-Noise Decomposition. Firstly, each gray-level image of the sequence will be decomposed on three components, Structure, Texture and Noise. The Structure and Texture components of each image of the sequence are taken to generate the background model. The absolute difference used to subtract the background before compute the binary image of moving objects. We, also, propose a video summarization based on the background subtraction results. The generated background model is used to compute the change during all time of the sequence. The experimental results demonstrate that our approach is effective and accurate for moving objects detection and yields a good summarization of the video sequence. - 2019 IEEE.This publication was made by NPRP grant # NPRP8-140-2-065 from the Qatar National Research Fund (a member of the Qatar Foundation). The statements made herein are solely the responsibility of the authors.Scopu

    Performance Analysis of DCT and DWT Algorithms in Image Steganography

    Get PDF
    Frequency domain techniques such as Discrete Cosine Transform (DCT) and Discrete Wavelet Transform (DWT) ensures high accuracy when compared with the spatial domain techniques. Therefore, these image steganographic methods were evaluated using public datasets to compare the performance of DCT and DWT. After performing different tests using the datasets in each of the algorithms, a comparative analysis is made in terms of the Peak Signal-to-Noise Ratio (PSNR) metrics. The results indicate that the stego image generated after embedding the secret acquires high imperceptibility and robustness. The performance of the DWT algorithm is higher as compared to the DCT algorithm and the resultant images produced are very less prone to noise attacks. In DCT and DWT algorithms, the cover image will be split based on 8×8 pixel blocks and 2D DCT is applied on each pixel. The secret will be embedded inside DCT coefficient and inverse 2D DCT is applied to recover the secret. Therefore, these image steganographic techniques can be adopted to transfer the confidential messages in different sectors. In the future, other data hiding methods using deep learning could be implemented to increase the robustness and imperceptibility of covert messages
    corecore